58 research outputs found
Smart speaker design and implementation with biometric authentication and advanced voice interaction capability
Advancements in semiconductor technology have reduced dimensions and cost
while improving the performance and capacity of chipsets. In addition,
advancement in the AI frameworks and libraries brings possibilities to
accommodate more AI at the resource-constrained edge of consumer IoT devices.
Sensors are nowadays an integral part of our environment which provide
continuous data streams to build intelligent applications. An example could be
a smart home scenario with multiple interconnected devices. In such smart
environments, for convenience and quick access to web-based service and
personal information such as calendars, notes, emails, reminders, banking, etc,
users link third-party skills or skills from the Amazon store to their smart
speakers. Also, in current smart home scenarios, several smart home products
such as smart security cameras, video doorbells, smart plugs, smart carbon
monoxide monitors, and smart door locks, etc. are interlinked to a modern smart
speaker via means of custom skill addition. Since smart speakers are linked to
such services and devices via the smart speaker user's account. They can be
used by anyone with physical access to the smart speaker via voice commands. If
done so, the data privacy, home security and other aspects of the user get
compromised. Recently launched, Tensor Cam's AI Camera, Toshiba's Symbio,
Facebook's Portal are camera-enabled smart speakers with AI functionalities.
Although they are camera-enabled, yet they do not have an authentication scheme
in addition to calling out the wake-word. This paper provides an overview of
cybersecurity risks faced by smart speaker users due to lack of authentication
scheme and discusses the development of a state-of-the-art camera-enabled,
microphone array-based modern Alexa smart speaker prototype to address these
risks
Imbal-OL: Online Machine Learning from Imbalanced Data Streams in Real-world IoT
Typically a Neural Networks (NN) is trained on data
centers using historic datasets, then a C source file (model as a
char array) of the trained model is generated and flashed on IoT
devices. This standard process impedes the flexibility of billions of
deployed ML-powered devices as they cannot learn unseen/fresh
data patterns (static intelligence) and are impossible to adapt
to dynamic scenarios. Currently, to address this issue, Online
Machine Learning (OL) algorithms are deployed on IoT devices
that provide devices the ability to locally re-train themselves -
continuously updating the last few NN layers using unseen data
patterns encountered after deployment.
In OL, catastrophic forgetting is common when NNs are
trained using non-stationary data distribution. The majority of
recent work in the OL domain embraces the implicit assumption
that the distribution of local training data is balanced. But the
fact is, the sensor data streams in real-world IoT are severely
imbalanced and temporally correlated. This paper introduces
Imbal-OL, a resource-friendly technique that can be used as
an OL plugin to balance the size of classes in a range of data
streams. When Imbal-OL processed stream is used for OL, the
models can adapt faster to changes in the stream while parallelly
preventing catastrophic forgetting. Experimental evaluation of
Imbal-OL using CIFAR datasets over ResNet-18 demonstrates
its ability to deal with imperfect data streams, as it manages
to produce high-quality models even under challenging learning
setting
Enabling Machine Learning on the Edge using SRAM Conserving Efficient Neural Networks Execution Approach
Edge analytics refers to the application of data analytics and Machine Learning (ML) algorithms on IoT devices. The concept of edge analytics is gaining popularity due to its ability to perform AI-based analytics at the device level, enabling autonomous decision-making, without depending on the cloud. However, the majority of Internet of Things (IoT) devices are embedded systems with a low-cost microcontroller unit (MCU) or a small CPU as its brain, which often are incapable of handling complex ML algorithms.
In this paper, we propose an approach for the ecient execution of already deeply compressed, large neural networks (NNs) on tiny IoT devices. After optimizing NNs using state-of-the-art deep model compression methods, when the resultant models are executed by MCUs or small CPUs using the model execution sequence produced by our approach, higher levels of conserved SRAM can be achieved. During the evaluation for nine popular models, when comparing the default NN execution sequence with the sequence produced by our approach, we found that 1.61-38.06% less SRAM was used to produce inference results, the inference time was reduced by 0.28-4.9 ms, and energy consumption was reduced by 4-84 mJ. Despite achieving such high conserved levels of SRAM, our method 100% preserved the accuracy, F1 score, etc. (model performance)
An Ontological Architecture for Principled and Automated System of Systems Composition
A distributed system's functionality must continuously evolve, especially when environmental context changes. Such required evolution imposes unbearable complexity on system development. An alternative is to make systems able to self-adapt by opportunistically composing at runtime to generate systems of systems (SoSs) that offer value-added functionality. The success of such an approach calls for abstracting the heterogeneity of systems and enabling the programmatic construction of SoSs with minimal developer intervention. We propose a general ontology-based approach to describe distributed systems, seeking to achieve abstraction and enable runtime reasoning between systems. We also propose an architecture for systems that utilize such ontologies to enable systems to discover and `understand' each other, and potentially compose, all at runtime. We detail features of the ontology and the architecture through two contrasting case studies. We also quantitatively evaluate the scalability and validity of our approach through experiments and simulations. Our approach enables system developers to focus on high-level SoS composition without being tied down with the specific deployment-specific implementation details
OTA-TinyML: Over the air deployment of TinyML models and execution on IoT devices
This article presents a novel over-the-air (OTA) technique to remotely deploy tiny ML models over Internet of Things (IoT) devices and perform tasks, such as machine learning (ML) model updates, firmware reflashing, reconfiguration, or repurposing. We discuss relevant challenges for OTA ML deployment over IoT both at the scientific and engineering level. We propose OTA-TinyML to enable resource-constrained IoT devices to perform end-to-end fetching, storage, and execution of many TinyML models. OTA-TinyML loads the C source file of ML models from a web server into the embedded IoT devices via HTTPS. OTA-TinyML is tested by performing remote fetching of six types of ML models, storing them on four types of memory units, then loading and executing on seven popular MCU boards
- …